35 research outputs found
Forward-Backward Stochastic Neural Networks: Deep Learning of High-dimensional Partial Differential Equations
Classical numerical methods for solving partial differential equations suffer
from the curse dimensionality mainly due to their reliance on meticulously
generated spatio-temporal grids. Inspired by modern deep learning based
techniques for solving forward and inverse problems associated with partial
differential equations, we circumvent the tyranny of numerical discretization
by devising an algorithm that is scalable to high-dimensions. In particular, we
approximate the unknown solution by a deep neural network which essentially
enables us to benefit from the merits of automatic differentiation. To train
the aforementioned neural network we leverage the well-known connection between
high-dimensional partial differential equations and forward-backward stochastic
differential equations. In fact, independent realizations of a standard
Brownian motion will act as training data. We test the effectiveness of our
approach for a couple of benchmark problems spanning a number of scientific
domains including Black-Scholes-Barenblatt and Hamilton-Jacobi-Bellman
equations, both in 100-dimensions
Parametric Gaussian Process Regression for Big Data
This work introduces the concept of parametric Gaussian processes (PGPs),
which is built upon the seemingly self-contradictory idea of making Gaussian
processes parametric. Parametric Gaussian processes, by construction, are
designed to operate in "big data" regimes where one is interested in
quantifying the uncertainty associated with noisy data. The proposed
methodology circumvents the well-established need for stochastic variational
inference, a scalable algorithm for approximating posterior distributions. The
effectiveness of the proposed approach is demonstrated using an illustrative
example with simulated data and a benchmark dataset in the airline industry
with approximately 6 million records
Deep Multi-fidelity Gaussian Processes
We develop a novel multi-fidelity framework that goes far beyond the
classical AR(1) Co-kriging scheme of Kennedy and O'Hagan (2000). Our method can
handle general discontinuous cross-correlations among systems with different
levels of fidelity. A combination of multi-fidelity Gaussian Processes (AR(1)
Co-kriging) and deep neural networks enables us to construct a method that is
immune to discontinuities. We demonstrate the effectiveness of the new
technology using standard benchmark problems designed to resemble the outputs
of complicated high- and low-fidelity codes
A multi-fidelity stochastic collocation method using locally improved reduced-order models
Over the last few years there have been dramatic advances in our
understanding of mathematical and computational models of complex systems in
the presence of uncertainty. This has led to a growth in the area of
uncertainty quantification as well as the need to develop efficient, scalable,
stable and convergent computational methods for solving differential equations
with random inputs. Stochastic Galerkin methods based on polynomial chaos
expansions have shown superiority to other non-sampling and many sampling
techniques. However, for complicated governing equations numerical
implementations of stochastic Galerkin methods can become non-trivial. On the
other hand, Monte Carlo and other traditional sampling methods, are
straightforward to implement. However, they do not offer as fast convergence
rates as stochastic Galerkin. Other numerical approaches are the stochastic
collocation (SC) methods, which inherit both, the ease of implementation of
Monte Carlo and the robustness of stochastic Galerkin to a great deal. However,
stochastic collocation and its powerful extensions, e.g. sparse grid stochastic
collocation, can simply fail to handle more levels of complication. The
seemingly innocent Burgers equation driven by Brownian motion is such an
example. In this work we propose a novel enhancement to stochastic collocation
methods using locally improved deterministic model reduction techniques that
can handle this pathological example and hopefully other more complicated
equations like Stochastic Navier-Stokes. Local improvements to reduced-order
models are achieved using sensitivity analysis of the proper orthogonal
decomposition. Our numerical results show that the proposed technique is not
only reliable and robust but also very efficient
Hidden Physics Models: Machine Learning of Nonlinear Partial Differential Equations
While there is currently a lot of enthusiasm about "big data", useful data is
usually "small" and expensive to acquire. In this paper, we present a new
paradigm of learning partial differential equations from {\em small} data. In
particular, we introduce \emph{hidden physics models}, which are essentially
data-efficient learning machines capable of leveraging the underlying laws of
physics, expressed by time dependent and nonlinear partial differential
equations, to extract patterns from high-dimensional data generated from
experiments. The proposed methodology may be applied to the problem of
learning, system identification, or data-driven discovery of partial
differential equations. Our framework relies on Gaussian processes, a powerful
tool for probabilistic inference over functions, that enables us to strike a
balance between model complexity and data fitting. The effectiveness of the
proposed approach is demonstrated through a variety of canonical problems,
spanning a number of scientific domains, including the Navier-Stokes,
Schr\"odinger, Kuramoto-Sivashinsky, and time dependent linear fractional
equations. The methodology provides a promising new direction for harnessing
the long-standing developments of classical methods in applied mathematics and
mathematical physics to design learning machines with the ability to operate in
complex domains without requiring large quantities of data
Deep Hidden Physics Models: Deep Learning of Nonlinear Partial Differential Equations
A long-standing problem at the interface of artificial intelligence and
applied mathematics is to devise an algorithm capable of achieving human level
or even superhuman proficiency in transforming observed data into predictive
mathematical models of the physical world. In the current era of abundance of
data and advanced machine learning capabilities, the natural question arises:
How can we automatically uncover the underlying laws of physics from
high-dimensional data generated from experiments? In this work, we put forth a
deep learning approach for discovering nonlinear partial differential equations
from scattered and potentially noisy observations in space and time.
Specifically, we approximate the unknown solution as well as the nonlinear
dynamics by two deep neural networks. The first network acts as a prior on the
unknown solution and essentially enables us to avoid numerical differentiations
which are inherently ill-conditioned and unstable. The second network
represents the nonlinear dynamics and helps us distill the mechanisms that
govern the evolution of a given spatiotemporal data-set. We test the
effectiveness of our approach for several benchmark problems spanning a number
of scientific domains and demonstrate how the proposed framework can help us
accurately learn the underlying dynamics and forecast future states of the
system. In particular, we study the Burgers', Korteweg-de Vries (KdV),
Kuramoto-Sivashinsky, nonlinear Schr\"{o}dinger, and Navier-Stokes equations
Deep Learning of Turbulent Scalar Mixing
Based on recent developments in physics-informed deep learning and deep
hidden physics models, we put forth a framework for discovering turbulence
models from scattered and potentially noisy spatio-temporal measurements of the
probability density function (PDF). The models are for the conditional expected
diffusion and the conditional expected dissipation of a Fickian scalar
described by its transported single-point PDF equation. The discovered model
are appraised against exact solution derived by the amplitude mapping closure
(AMC)/ Johnsohn-Edgeworth translation (JET) model of binary scalar mixing in
homogeneous turbulence.Comment: arXiv admin note: text overlap with arXiv:1808.04327,
arXiv:1808.0895
Numerical Gaussian Processes for Time-dependent and Non-linear Partial Differential Equations
We introduce the concept of numerical Gaussian processes, which we define as
Gaussian processes with covariance functions resulting from temporal
discretization of time-dependent partial differential equations. Numerical
Gaussian processes, by construction, are designed to deal with cases where: (1)
all we observe are noisy data on black-box initial conditions, and (2) we are
interested in quantifying the uncertainty associated with such noisy data in
our solutions to time-dependent partial differential equations. Our method
circumvents the need for spatial discretization of the differential operators
by proper placement of Gaussian process priors. This is an attempt to construct
structured and data-efficient learning machines, which are explicitly informed
by the underlying physics that possibly generated the observed data. The
effectiveness of the proposed approach is demonstrated through several
benchmark problems involving linear and nonlinear time-dependent operators. In
all examples, we are able to recover accurate approximations of the latent
solutions, and consistently propagate uncertainty, even in cases involving very
long time integration
Inferring solutions of differential equations using noisy multi-fidelity data
For more than two centuries, solutions of differential equations have been
obtained either analytically or numerically based on typically well-behaved
forcing and boundary conditions for well-posed problems. We are changing this
paradigm in a fundamental way by establishing an interface between
probabilistic machine learning and differential equations. We develop
data-driven algorithms for general linear equations using Gaussian process
priors tailored to the corresponding integro-differential operators. The only
observables are scarce noisy multi-fidelity data for the forcing and solution
that are not required to reside on the domain boundary. The resulting
predictive posterior distributions quantify uncertainty and naturally lead to
adaptive solution refinement via active learning. This general framework
circumvents the tyranny of numerical discretization as well as the consistency
and stability issues of time-integration, and is scalable to high-dimensions.Comment: 19 pages, 3 figure
Hidden Fluid Mechanics: A Navier-Stokes Informed Deep Learning Framework for Assimilating Flow Visualization Data
We present hidden fluid mechanics (HFM), a physics informed deep learning
framework capable of encoding an important class of physical laws governing
fluid motions, namely the Navier-Stokes equations. In particular, we seek to
leverage the underlying conservation laws (i.e., for mass, momentum, and
energy) to infer hidden quantities of interest such as velocity and pressure
fields merely from spatio-temporal visualizations of a passive scaler (e.g.,
dye or smoke), transported in arbitrarily complex domains (e.g., in human
arteries or brain aneurysms). Our approach towards solving the aforementioned
data assimilation problem is unique as we design an algorithm that is agnostic
to the geometry or the initial and boundary conditions. This makes HFM highly
flexible in choosing the spatio-temporal domain of interest for data
acquisition as well as subsequent training and predictions. Consequently, the
predictions made by HFM are among those cases where a pure machine learning
strategy or a mere scientific computing approach simply cannot reproduce. The
proposed algorithm achieves accurate predictions of the pressure and velocity
fields in both two and three dimensional flows for several benchmark problems
motivated by real-world applications. Our results demonstrate that this
relatively simple methodology can be used in physical and biomedical problems
to extract valuable quantitative information (e.g., lift and drag forces or
wall shear stresses in arteries) for which direct measurements may not be
possible